11 research outputs found

    Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme

    Full text link
    Joint extraction of entities and relations is an important task in information extraction. To tackle this problem, we firstly propose a novel tagging scheme that can convert the joint extraction task to a tagging problem. Then, based on our tagging scheme, we study different end-to-end models to extract entities and their relations directly, without identifying entities and relations separately. We conduct experiments on a public dataset produced by distant supervision method and the experimental results show that the tagging based methods are better than most of the existing pipelined and joint learning methods. What's more, the end-to-end model proposed in this paper, achieves the best results on the public dataset

    TAG : Type Auxiliary Guiding for Code Comment Generation

    Full text link
    Existing leading code comment generation approaches with the structure-to-sequence framework ignores the type information of the interpretation of the code, e.g., operator, string, etc. However, introducing the type information into the existing framework is non-trivial due to the hierarchical dependence among the type information. In order to address the issues above, we propose a Type Auxiliary Guiding encoder-decoder framework for the code comment generation task which considers the source code as an N-ary tree with type information associated with each node. Specifically, our framework is featured with a Type-associated Encoder and a Type-restricted Decoder which enables adaptive summarization of the source code. We further propose a hierarchical reinforcement learning method to resolve the training difficulties of our proposed framework. Extensive evaluations demonstrate the state-of-the-art performance of our framework with both the auto-evaluated metrics and case studies.Comment: ACL 2020, Accepte

    Implementation of Lightweight Convolutional Neural Networks via Layer-Wise Differentiable Compression

    No full text
    Convolutional neural networks (CNNs) have achieved significant breakthroughs in various domains, such as natural language processing (NLP), and computer vision. However, performance improvement is often accompanied by large model size and computation costs, which make it not suitable for resource-constrained devices. Consequently, there is an urgent need to compress CNNs, so as to reduce model size and computation costs. This paper proposes a layer-wise differentiable compression (LWDC) algorithm for compressing CNNs structurally. A differentiable selection operator OS is embedded in the model to compress and train the model simultaneously by gradient descent in one go. Instead of pruning parameters from redundant operators by contrast to most of the existing methods, our method replaces the original bulky operators with more lightweight ones directly, which only needs to specify the set of lightweight operators and the regularization factor in advance, rather than the compression rate for each layer. The compressed model produced by our method is generic and does not need any special hardware/software support. Experimental results on CIFAR-10, CIFAR-100 and ImageNet have demonstrated the effectiveness of our method. LWDC obtains more significant compression than state-of-the-art methods in most cases, while having lower performance degradation. The impact of lightweight operators and regularization factor on the compression rate and accuracy also is evaluated

    Attention Round for Post-Training Quantization

    Full text link
    At present, the quantification methods of neural network models are mainly divided into post-training quantization (PTQ) and quantization aware training (QAT). Post-training quantization only need a small part of the data to complete the quantification process, but the performance of its quantitative model is not as good as the quantization aware training. This paper presents a novel quantification method called Attention Round. This method gives parameters w the opportunity to be mapped to all possible quantized values, rather than just the two quantized values nearby w in the process of quantization. The probability of being mapped to different quantified values is negatively correlated with the distance between the quantified values and w, and decay with a Gaussian function. In addition, this paper uses the lossy coding length as a measure to assign bit widths to the different layers of the model to solve the problem of mixed precision quantization, which effectively avoids to solve combinatorial optimization problem. This paper also performs quantitative experiments on different models, the results confirm the effectiveness of the proposed method. For ResNet18 and MobileNetV2, the post-training quantization proposed in this paper only require 1,024 training data and 10 minutes to complete the quantization process, which can achieve quantization performance on par with quantization aware training.Comment: 18 pages, 5 figures, 5 table

    LA-ICP-MS U-Pb geochronology and clumped isotope constraints on the formation and evolution of an ancient dolomite reservoir: The Middle Permian of northwest Sichuan Basin (SW China)

    No full text
    Recent advances in laser-ablation inductively coupled-plasma mass spectrometry (LA-ICP-MS) in-situ U-Pb radiometric dating and clumped isotope thermometry (Δ) of carbonate minerals provide potential for refining the fluid flow and diagenetic history of carbonate successions. In this study, the Middle Permian dolomites proximal to the Longmenshan fold and thrust belt in northwest Sichuan Basin, southwest China, were investigated using combined U-Pb geochronology, clumped isotope thermometry, and routine isotopic (δC, δO and Sr/Sr) and elemental geochemistry, in an attempt to reveal the possible relevance of carbonate diagenesis and porosity evolution to the basin-scale tectonic/fluid flow events in the framework of absolute time. Overall, formation and evolution of these dolomites temporally correlated well with major tectonic episodes of the Longmenshan fold and thrust belt. Contrary to the previously-assumed volcanic-related model, U-Pb dating and Δ analyses suggest a mid- to late-Triassic replacive dolomitization event (U-Pb ages of 240 ± 12 Ma to 233.8 ± 6.4 Ma) by a hot (Δ temperatures 88– 104 °C) basinal brine, which was likely driven by thrust-related compression of the Longmenshan fold and thrust belt during the Triassic. Replacive dolomitization was immediately followed by cementation of euhedral dolomites (U-Pb age of 228 ± 10 Ma) and blocky calcites (U-Pb ages of 224.8 ± 1.8 Ma to 213.4 ± 3.3 Ma) precipitated from the basinal brine in the same tectonic regime. Afterwards, there was long-term cessation of diagenesis that was coincident with, and thus may well be attributed to, tectono-thermal quiescence during most of the Mesozoic Era and Paleogene. Finally, saddle dolomite cements yield Miocene ages (U-Pb ages of 16.40 ± 0.74 Ma to 12.3 ± 1.2 Ma) and precipitated in pre-existing vugs, representing a hydrothermal fluid flow event (Δ temperatures up to 170 °C) during the reactivation of thrusting. In addition, similarities in U-Pb ages and geochemical signatures between host rocks (replacive dolomites) and the vug-lining dolomite cements suggest that most porosity in these dolomite reservoirs was inherited from the precursor limestones, which probably experienced meteoric leaching during late Permian uplift. This study highlights that there are geodynamic controls on fluid flow and consequent diagenetic evolution of carbonates in tectonically active sedimentary basins. It illustrates that burial alone in this setting is not a sufficient driver for diagenetic alteration and porosity reduction. The combination of U-Pb radiometric dating, clumped isotope thermometry and routine geochemical analyses is a useful approach in refining the diagenetic and porosity evolution history of ancient carbonate successions
    corecore